Adaptive Knowledge Driven Regularization for Deep Neural Networks

نویسندگان

چکیده

In many real-world applications, the amount of data available for training is often limited, and thus inductive bias auxiliary knowledge are much needed regularizing model training. One popular regularization method to impose prior distribution assumptions on parameters, recent works also attempt regularize by integrating external into specific neurons. However, existing methods did not take account interaction between connected neuron pairs, which invaluable internal adaptive better representation learning as progresses. this paper, we explicitly neurons, propose an driven method, CORR-Reg. The key idea CORR-Reg give a higher significance weight connections more correlated pairs. weights adaptively identify important input neurons each neuron. Instead connection parameters with static strength such decay, imposes weaker significant connections. As consequence, attend informative features learn diversified discriminative representation. We derive Bayesian inference framework novel optimization algorithm Lagrange multiplier Stochastic Gradient Descent. Extensive evaluations diverse benchmark datasets neural network structures show that achieves improvement over state-of-the-art methods.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Group sparse regularization for deep neural networks

In this paper, we consider the joint task of simultaneously optimizing (i) the weights of a deep neural network, (ii) the number of neurons for each hidden layer, and (iii) the subset of active input features (i.e., feature selection). While these problems are generally dealt with separately, we present a simple regularized formulation allowing to solve all three of them in parallel, using stan...

متن کامل

Adaptive dropout for training deep neural networks

Recently, it was shown that deep neural networks can perform very well if the activities of hidden units are regularized during learning, e.g, by randomly dropping out 50% of their activities. We describe a method called ‘standout’ in which a binary belief network is overlaid on a neural network and is used to regularize of its hidden units by selectively setting activities to zero. This ‘adapt...

متن کامل

Knowledge Projection for Deep Neural Networks

While deeper and wider neural networks are actively pushing the performance limits of various computer vision and machine learning tasks, they often require large sets of labeled data for effective training and suffer from extremely high computational complexity. In this paper, we will develop a new framework for training deep neural networks on datasets with limited labeled samples using cross...

متن کامل

Stochastic Pooling for Regularization of Deep Convolutional Neural Networks

We introduce a simple and effective method for regularizing large convolutional neural networks. We replace the conventional deterministic pooling operations with a stochastic procedure, randomly picking the activation within each pooling region according to a multinomial distribution, given by the activities within the pooling region. The approach is hyper-parameter free and can be combined wi...

متن کامل

A constrained regularization approach for input-driven recurrent neural networks

We introduce a novel regularization approach for a class of inputdriven recurrent neural networks. The regularization of network parameters is constrained to reimplement a previously recorded state trajectory. We derive a closed-form solution for network regularization and show that the method is capable of reimplementing harvested dynamics. We investigate important properties of the method and...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i10.17067